Deep learning models can achieve high accuracy when trained on large amounts of labeled data. However, real-world scenarios often involve several challenges: Training data may become available in installments, may originate from multiple different domains, and may not contain labels for training. Certain settings, for instance medical applications, often involve further restrictions that prohibit retention of previously seen data due to privacy regulations. In this work, to address such challenges, we study unsupervised segmentation in continual learning scenarios that involve domain shift. To that end, we introduce GarDA (Generative Appearance Replay for continual Domain Adaptation), a generative-replay based approach that can adapt a segmentation model sequentially to new domains with unlabeled data. In contrast to single-step unsupervised domain adaptation (UDA), continual adaptation to a sequence of domains enables leveraging and consolidation of information from multiple domains. Unlike previous approaches in incremental UDA, our method does not require access to previously seen data, making it applicable in many practical scenarios. We evaluate GarDA on two datasets with different organs and modalities, where it substantially outperforms existing techniques.
translated by 谷歌翻译
多个实例学习(MIL)方法在数字病理学中对GIGA像素大小的全型图像(WSI)进行分类变得越来越流行。大多数MIL方法通过处理所有组织斑块,以单个WSI放大倍率运行。这样的公式诱导了高计算要求,并将WSI级表示的上下文化限制为单个量表。一些MIL方法扩展到多个量表,但在计算上要求更高。在本文中,受病理诊断过程的启发,我们提出了Zoommil,该方法学会了以端到端的方式执行多层缩放。Zoommil通过从多个增强元中汇总组织信息来构建WSI表示。所提出的方法在两个大数据集上的WSI分类中优于最先进的MIL方法,同时大大降低了关于浮点操作(FLOPS)和处理时间的计算需求,最高为40倍。
translated by 谷歌翻译
3D光学相干断层扫描图像中视网膜流体的准确分割是诊断和个性化眼部疾病的关键。尽管深度学习在这项任务上取得了成功,但受过训练的监督模型通常会因不像标记示例的图像而失败,例如对于使用不同设备获取的图像。我们在此提出了一个新型的半监督学习框架,用于从新未标记的域分割体积图像。我们共同使用受监督和对比度学习,还引入了一种对比配对方案,该方案利用3D中附近切片之间的相似性。此外,我们建议通过渠道聚合作为对比特征图投影的常规空间释放聚合的替代方法。我们评估了从(标记的)源域对(未标记的)目标域的域适应方法,每个方法都包含具有不同采集设备的图像。在目标域中,我们的方法获得了比SIMCLR(最先进的对比框架)高13.8%的骰子系数,并导致结果可与该领域中有监督的训练的上限相当。在源域中,我们的模型还通过成功利用来自许多未标记的图像的信息,将结果提高了5.4%。
translated by 谷歌翻译
Neural networks are prone to catastrophic forgetting when trained incrementally on different tasks. Popular incremental learning methods mitigate such forgetting by retaining a subset of previously seen samples and replaying them during the training on subsequent tasks. However, this is not always possible, e.g., due to data protection regulations. In such restricted scenarios, one can employ generative models to replay either artificial images or hidden features to a classifier. In this work, we propose Genifer (GENeratIve FEature-driven image Replay), where a generative model is trained to replay images that must induce the same hidden features as real samples when they are passed through the classifier. Our technique therefore incorporates the benefits of both image and feature replay, i.e.: (1) unlike conventional image replay, our generative model explicitly learns the distribution of features that are relevant for classification; (2) in contrast to feature replay, our entire classifier remains trainable; and (3) we can leverage image-space augmentations, which increase distillation performance while also mitigating overfitting during the training of the generative model. We show that Genifer substantially outperforms the previous state of the art for various settings on the CIFAR-100 and CUB-200 datasets.
translated by 谷歌翻译
Tracking objects over long videos effectively means solving a spectrum of problems, from short-term association for un-occluded objects to long-term association for objects that are occluded and then reappear in the scene. Methods tackling these two tasks are often disjoint and crafted for specific scenarios, and top-performing approaches are often a mix of techniques, which yields engineering-heavy solutions that lack generality. In this work, we question the need for hybrid approaches and introduce SUSHI, a unified and scalable multi-object tracker. Our approach processes long clips by splitting them into a hierarchy of subclips, which enables high scalability. We leverage graph neural networks to process all levels of the hierarchy, which makes our model unified across temporal scales and highly general. As a result, we obtain significant improvements over state-of-the-art on four diverse datasets. Our code and models will be made available.
translated by 谷歌翻译
图提供了一种自然的方式来制定多个对象跟踪(MOT)和多个对象跟踪和分割(MOTS),逐个检测范式中。但是,他们还引入了学习方法的主要挑战,因为定义可以在这种结构化领域运行的模型并不是微不足道的。在这项工作中,我们利用MOT的经典网络流程公式来定义基于消息传递网络(MPN)的完全微分框架。通过直接在图形域上操作,我们的方法可以在整个检测和利用上下文特征上全球推理。然后,它共同预测了数据关联问题的最终解决方案和场景中所有对象的分割掩码,同时利用这两个任务之间的协同作用。我们在几个公开可用的数据集中获得跟踪和细分的最新结果。我们的代码可在github.com/ocetintas/mpntrackseg上找到。
translated by 谷歌翻译
Guillain-Barre综合征是一种罕见的神经系统疾病,其中人免疫系统攻击周围神经系统。周围神经系统似乎是神经元模型的数学模型的扩散连接系统,并且该系统的周期比每个神经回路的周期都短。传导路径中的刺激将被轴突接收到失去其功能的髓鞘鞘,并在外部传递到靶器官,旨在解决降低神经传导的问题。在神经元模拟环境中,可以创建神经元模型并定义系统内发生的生物物理事件。在这种环境中,细胞和树突之间的信号传递是图形的。模拟的钾和钠电导是充分复制的,电子动作电位与实验测量的电位相当。在这项工作中,我们提出了一个模拟和数字耦合的神经元模型,该模型包括个人兴奋性和抑制性神经回路块,用于低成本和节能系统。与数字设计相比,我们的模拟设计的性能较低,但能源效率降低了32.3 \%。因此,所得的耦合模拟硬件神经元模型可以是模拟神经传导减少的模型。结果,模拟耦合的神经元(即使具有更大的设计复杂性)为未来开发的可穿戴传感器设备的竞争者,该设备可能有助于治疗吉兰 - 巴雷综合症和其他神经系统疾病。
translated by 谷歌翻译